Learning to Pick at Non-Zero-Velocity From Interactive Demonstrations

نویسندگان

چکیده

This work investigates how the intricate task of a continuous pick & place (P&P) motion may be learned from humans based on demonstrations and corrections. Due to complexity task, these are often slow even slightly flawed, particularly at moments when multiple aspects (i.e., end-effector movement, orientation, gripper width) have demonstrated once. Rather than training person give better demonstrations, non-expert users provided with ability interactively modify dynamics their initial demonstration through teleoperated corrective feedback. in turn allows them teach motions outside own physical capabilities. In end, goal is obtain faster but reliable execution task. The presented framework learns desired movement current Cartesian position Gaussian Processes (GPs), resulting reactive, time-invariant policy. Using GPs also online interactive corrections active disturbance rejection epistemic uncertainty minimization. experimental evaluation carried out Franka-Emika Panda. Tests were performed determine i) framework’s effectiveness successfully learning quickly an object, ii) ease policy correction environmental changes different object sizes mass), iii) usability for users.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Non-interactive Zero-Knowledge from Homomorphic Encryption

We propose a method for compiling a class of Σ-protocols (3-move public-coin protocols) into non-interactive zero-knowledge arguments. The method is based on homomorphic encryption and does not use random oracles. It only requires that a private/public key pair is set up for the verifier. The method applies to all known discrete-log based Σ-protocols. As applications, we obtain non-interactive ...

متن کامل

Learning from Limited Demonstrations

We propose a Learning from Demonstration (LfD) algorithm which leverages expert data, even if they are very few or inaccurate. We achieve this by using both expert data, as well as reinforcement signals gathered through trial-and-error interactions with the environment. The key idea of our approach, Approximate Policy Iteration with Demonstration (APID), is that expert’s suggestions are used to...

متن کامل

Learning from Demonstrations Through the Use of Non-rigid Registration

We consider the problem of teaching robots by demonstration how to perform manipulation tasks, in which the geometry (including size, shape, and pose) of the relevant objects varies from trial to trial. We present a method, which we call trajectory transfer, for adapting a demonstrated trajectory from the geometry at training time to the geometry at test time. Trajectory transfer is based on no...

متن کامل

Robot Learning from Failed Demonstrations

Robot learning from demonstration (RLfD) seeks to enable lay users to encode desired robot behaviors as autonomous controllers. Current work uses a human’s demonstration of the target task to initialize the robot’s policy, and then improves its performance either through practice (with a known reward function), or additional human interaction. In this article, we focus on the initialization ste...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE robotics and automation letters

سال: 2022

ISSN: ['2377-3766']

DOI: https://doi.org/10.1109/lra.2022.3165531